GRADES: Gradient Descent for Similarity Caching
نویسندگان
چکیده
A similarity cache can reply to a query for an object with similar objects stored locally. In some applications of caches, queries and are naturally represented as points in continuous space. This is example the case $360^\circ$ videos where user’s head orientation—expressed spherical coordinates—determines what part video needs be retrieved, or recommendation systems metric learning technique used embed finite dimensional space opportune distance capture content dissimilarity. Existing caching policies simple modifications classic like LRU, LFU, qLRU ignore nature embedded. this paper, we propose Grades, new policy that uses gradient descent navigate find appropriate store cache. We provide theoretical convergence guarantees show Grades increases served by both mentioned above.
منابع مشابه
Handwritten Character Recognition using Modified Gradient Descent Technique of Neural Networks and Representation of Conjugate Descent for Training Patterns
The purpose of this study is to analyze the performance of Back propagation algorithm with changing training patterns and the second momentum term in feed forward neural networks. This analysis is conducted on 250 different words of three small letters from the English alphabet. These words are presented to two vertical segmentation programs which are designed in MATLAB and based on portions (1...
متن کاملLearning to learn by gradient descent by gradient descent
The move from hand-designed features to learned features in machine learning has been wildly successful. In spite of this, optimization algorithms are still designed by hand. In this paper we show how the design of an optimization algorithm can be cast as a learning problem, allowing the algorithm to learn to exploit structure in the problems of interest in an automatic way. Our learned algorit...
متن کاملEmpirical Comparison of Gradient Descent andExponentiated Gradient Descent in
This report describes a series of results using the exponentiated gradient descent (EG) method recently proposed by Kivinen and Warmuth. Prior work is extended by comparing speed of learning on a nonstationary problem and on an extension to backpropagation networks. Most signi cantly, we present an extension of the EG method to temporal-di erence and reinforcement learning. This extension is co...
متن کاملAccelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent
Nesterov's accelerated gradient descent (AGD), an instance of the general family of"momentum methods", provably achieves faster convergence rate than gradient descent (GD) in the convex setting. However, whether these methods are superior to GD in the nonconvex setting remains open. This paper studies a simple variant of AGD, and shows that it escapes saddle points and finds a second-order stat...
متن کاملLearning to Learn without Gradient Descent by Gradient Descent
We learn recurrent neural network optimizers trained on simple synthetic functions by gradient descent. We show that these learned optimizers exhibit a remarkable degree of transfer in that they can be used to efficiently optimize a broad range of derivative-free black-box functions, including Gaussian process bandits, simple control objectives, global optimization benchmarks and hyper-paramete...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE ACM Transactions on Networking
سال: 2023
ISSN: ['1063-6692', '1558-2566']
DOI: https://doi.org/10.1109/tnet.2022.3187044